Nonparametric Information Geometry: From Divergence Function to Referential-Representational Biduality on Statistical Manifolds

نویسنده

  • Jun Zhang
چکیده

Divergence functions are the non-symmetric “distance” on the manifold,Mθ, of parametric probability density functions over a measure space, (X,μ). Classical information geometry prescribes, on Mθ: (i) a Riemannian metric given by the Fisher information; (ii) a pair of dual connections (giving rise to the family of α-connections) that preserve the metric under parallel transport by their joint actions; and (iii) a family of divergence functions (α-divergence) defined on Mθ × Mθ, which induce the metric and the dual connections. Here, we construct an extension of this differential geometric structure from Mθ (that of parametric probability density functions) to the manifold,M, of non-parametric functions on X , removing the positivity and normalization constraints. The generalized Fisher information and α-connections on M are induced by an α-parameterized family of divergence functions, reflecting the fundamental convex inequality associated with any smooth and strictly convex function. The infinite-dimensional manifold, M, has zero curvature for all these α-connections; hence, the generally non-zero curvature ofMθ can be interpreted as arising from an embedding ofMθ intoM. Furthermore, when a parametric model (after a monotonic scaling) forms an affine submanifold, its natural and expectation parameters form biorthogonal coordinates, and such a submanifold is dually flat for α = ±1, generalizing the results of Amari’s α-embedding. The present analysis illuminates two different types of duality in information geometry, one concerning the referential status of a point (measurable function) expressed in the divergence function (“referential duality”) and the other concerning its representation under an arbitrary monotone scaling (“representational duality”). Entropy 2013, 15 5385

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Reference Duality and Representation Duality in Information Geometry

Classical information geometry prescribes, on the parametric family of probability functions Mθ : (i) a Riemannian metric given by the Fisher information; (ii) a pair of dual connections (giving rise to the family of α-connections) that preserve the metric under parallel transport by their joint actions; and (iii) a family of (non-symmetric) divergence functions (α-divergence) defined on Mθ ×Mθ...

متن کامل

Divergence Function, Duality, and Convex Analysis

From a smooth, strictly convex function phi: Rn --> R, a parametric family of divergence function Dphi(alpha) may be introduced: [ equation: see text] for x, y epsilon int dom (Phi) subset Rn, and for alpha in R, with Dphi(+/-1) defined through taking the limit of alpha. Each member is shown to induce an alpha-independent Riemannian metric, as well as a pair of dual alpha-connections, which are...

متن کامل

Symplectic and Kähler Structures on Statistical Manifolds Induced from Divergence Functions

Divergence functions play a central role in information geometry. Given a manifold M, a divergence function D is a smooth, nonnegative function on the product manifold M×M that achieves its global minimum of zero (with semi-positive definite Hessian) at those points that form its diagonal submanifold Mx. It is well-known (Eguchi, 1982) that the statistical structure on M (a Riemmanian metric wi...

متن کامل

Examples of the Application of Nonparametric Information Geometry to Statistical Physics

We review a nonparametric version of Amari’s information geometry in which the set of positive probability densities on a given sample space is endowed with an atlas of charts to form a differentiable manifold modeled on Orlicz Banach spaces. This nonparametric setting is used to discuss the setting of typical problems in machine learning and statistical physics, such as black-box optimization,...

متن کامل

A Geometry Preserving Kernel over Riemannian Manifolds

Abstract- Kernel trick and projection to tangent spaces are two choices for linearizing the data points lying on Riemannian manifolds. These approaches are used to provide the prerequisites for applying standard machine learning methods on Riemannian manifolds. Classical kernels implicitly project data to high dimensional feature space without considering the intrinsic geometry of data points. ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Entropy

دوره 15  شماره 

صفحات  -

تاریخ انتشار 2013